Asymptotic Properties of Nearest Neighbor Rules Using Edited Data

نویسنده

  • Dennis L. Wilson
چکیده

The convergence properties of a nearest neighbor rule that uses an editing procedure to reduce the number of preclassified samples and to improve the performance of the rule are developed. Editing of the preclassified samples using the three-nearest neighbor rule followed by classification using the single-nearest neighbor rule with the remaining preclassified samples appears to produce a decision procedure whose risk approaches the Bayes' risk quite closely in many problems with only a few preclassified samples. The asymptotic risk of the nearest neighbor rules and the nearest neighbor rules using edited preclassified samples is calculated for several problems.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Asymptotic Behaviors of Nearest Neighbor Kernel Density Estimator in Left-truncated Data

Kernel density estimators are the basic tools for density estimation in non-parametric statistics.  The k-nearest neighbor kernel estimators represent a special form of kernel density estimators, in  which  the  bandwidth  is varied depending on the location of the sample points. In this paper‎, we  initially introduce the k-nearest neighbor kernel density estimator in the random left-truncatio...

متن کامل

Considerations about sample-size sensitivity of a family of edited nearest-neighbor rules

The edited nearest neighbor classification rules constitute a valid alternative to k-NN rules and other nonparametric classifiers. Experimental results with synthetic and real data from various domains and from different researchers and practitioners suggest that some editing algorithms (especially, the optimal ones) are very sensitive to the total number of prototypes considered. This paper in...

متن کامل

Submitted to CVPR ' 99 Bayesian based Optimal Nearest Neighbor

Nearest neighbor rules are popular classiiers, partly due to their good asymptotic properties. Improving the performance of NN rules when only nite samples are available has been studied over the last two decades. For example, smart distance measures which are data dependent have been proposed. From a theoretical point of view, one is interested in deriving the optimal NN distance measure. Exis...

متن کامل

On the edited fuzzy K-nearest neighbor rule

Classification of objects is an important area in a variety of fields and applications. In the presence of full knowledge of the underlying joint distributions, Bayes analysis yields an optimal decision procedure and produces optimal error rates. Many different methods are available to make a decision in those cases where information of the underlying joint distributions is not presented. The k...

متن کامل

Application of Proximity Graphs to Editing Nearest Neighbor Decision Rules

Non-parametric decision rules, such as the nearest neighbor (NN) rule, are attractive because no a priori knowledge is required concerning the underlying distributions of the data. Two traditional criticisms directed at the NN-rule concern the large amounts of storage and computation involved due to the apparent necessity to store all the sample (training) data. Thus there has been considerable...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • IEEE Trans. Systems, Man, and Cybernetics

دوره 2  شماره 

صفحات  -

تاریخ انتشار 1972